# Enhanced Mathematical Reasoning
Qwen3 4B Llamafile
Apache-2.0
Qwen3-4B is the latest generation large language model in the Qwen series, featuring 4B parameters, supporting a 128k context window and over 100 languages, with outstanding performance in reasoning, instruction following, and agent capabilities.
Large Language Model
Q
Mozilla
995
2
Qwen3 235B A22B GGUF
Apache-2.0
Qwen3 is the latest version of the Tongyi Qianwen series of large language models, offering a complete suite of dense models and Mixture of Experts (MoE) models. Based on extensive training data, Qwen3 achieves breakthrough progress in reasoning capabilities, instruction following, agent functionalities, and multilingual support.
Large Language Model
Q
Qwen
1,576
2
Qwen3 4B AWQ
Apache-2.0
Qwen3-4B is the latest generation of the Qwen series' 4-billion-parameter large language model, supporting switching between reasoning and non-reasoning modes with powerful inference, instruction following, and agent capabilities.
Large Language Model
Transformers

Q
Qwen
2,755
4
Phi 4 Reasoning Plus
MIT
Phi-4 Reasoning Enhanced is a 14-billion parameter open-source reasoning model developed by Microsoft Research, optimized through supervised fine-tuning and reinforcement learning, focusing on advanced reasoning capabilities in mathematics, science, and programming fields.
Large Language Model
Transformers Supports Multiple Languages

P
unsloth
189
2
Qwen3 1.7B
Apache-2.0
Qwen3 is the latest generation of large language models in the Tongyi Qianwen series, offering a complete set of dense models and Mixture of Experts (MoE) models. Based on large-scale training, Qwen3 has achieved breakthroughs in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model
Transformers

Q
Qwen
395.72k
113
Qwen2.5 Math 7B 16k Think
MIT
An improved model based on Qwen2.5-Math-7B, with extended context window and optimized reasoning capabilities
Large Language Model
Transformers

Q
Elliott
3,496
1
DAPO Qwen 32B
Apache-2.0
A large language model trained using the DAPO algorithm based on the Qwen2.5-32B model, specializing in mathematical problem-solving and multilingual text generation
Large Language Model
Safetensors Supports Multiple Languages
D
BytedTsinghua-SIA
7,241
3
Reasonflux F1
Other
ReasonFlux-F1-32B is a hierarchical large language model based on template-enhanced reasoning trajectory fine-tuning, excelling in reasoning tasks.
Large Language Model
Transformers

R
Gen-Verse
123
8
Open RS1
MIT
A small-scale large language model enhanced by reinforcement learning, focused on improving the reasoning capabilities of a 1.5B parameter model
Large Language Model
Transformers

O
knoveleng
6,229
4
Sombrero QwQ 32B Elite11
Apache-2.0
A large language model optimized based on Qwen's QwQ 32B architecture, focusing on efficient memory utilization, programming assistance, and complex problem-solving.
Large Language Model
Transformers English

S
prithivMLmods
1,201
8
Li 14b V0.4 Slerp0.1
This is a 14B-parameter large language model merged using the SLERP method, combining two base models: li-14b-v0.4 and miscii-14b-0218.
Large Language Model
Transformers

L
wanlige
70
7
Phi 4 Model Stock V2
Phi-4-Model-Stock-v2 is a large language model merged from multiple Phi-4 variant models using the model_stock merging method, demonstrating strong performance across multiple benchmarks.
Large Language Model
Transformers

P
bunnycore
56
2
Luxllama
Apache-2.0
LuxLlama is a fine-tuned version of the Meta-Llama-3.1-8B-Instruct model, specifically optimized for Luxembourgish language understanding and generation, as well as general and mathematical reasoning capabilities.
Large Language Model
Transformers Supports Multiple Languages

L
aiplanet
33
2
Skywork O1 Open PRM Qwen 2.5 1.5B
Other
Skywork o1 Open-PRM-Qwen-2.5-1.5B is an incremental process reward model trained on Qwen2.5-Math-1.5B-Instruct, specifically designed to enhance small-scale complex problem-solving capabilities.
Large Language Model
PyTorch
S
Skywork
4,368
30
Skywork O1 Open PRM Qwen 2.5 7B
Other
A 7B-parameter model in the Skywork o1 open model series, trained on Qwen2.5-Math-7B-Instruct, featuring enhanced reasoning capabilities through progressive process rewards
Large Language Model
S
Skywork
3,608
50
Powerlm 3b
Apache-2.0
PowerLM-3B is a small language model with 3 billion parameters, trained using the Power learning rate scheduler, and demonstrates excellent performance across multiple benchmarks including natural language multiple-choice, code generation, and mathematical reasoning.
Large Language Model
Transformers

P
ibm-research
11.07k
20
Chicka Mixtral 3x7b
MIT
A Mixture of Experts large language model based on 3 Mistral architecture models, excelling in dialogue, code, and mathematical tasks
Large Language Model
Transformers

C
Chickaboo
77
3
Quietstar 8 Ahead
Based on the Mistral-7b model, it employs the Quiet-STaR method for continuous pretraining, generating 8 reasoning tokens before each output token to enhance reasoning capabilities.
Large Language Model
Transformers

Q
ezelikman
239
90
Smaug 34B V0.1
Apache-2.0
A large language model fine-tuned based on jondurbin/bagel-34b-v0.2, optimized with novel DPO-Positive (DPOP) technology for preference learning
Large Language Model
Transformers

S
abacusai
2,694
62
Openchat 3.5 0106
Apache-2.0
OpenChat is an open-source language model based on Mistral-7B, trained with mixed-quality data to deliver exceptional text generation capabilities.
Large Language Model
Transformers

O
openchat
85.06k
357
Openchat 3.5 1210
Apache-2.0
OpenChat 3.5 1210 is an open-source language model based on the Mistral-7B architecture, trained on mixed-quality datasets, supporting programming, general dialogue, and mathematical reasoning tasks, with overall performance surpassing ChatGPT (March version) and Grok-1.
Large Language Model
Transformers

O
openchat
53.23k
272
Discolm Mixtral 8x7b V2
Apache-2.0
Experimental 8x7b Mixture of Experts model developed based on Mistral AI's Mixtral 8x7b, fine-tuned on Synthia, MetaMathQA, and Capybara datasets
Large Language Model
Transformers English

D
DiscoResearch
205
124
Metamath Llemma 7B
Apache-2.0
MetaMath-Llemma-7B is a math-specialized large language model fully fine-tuned on the MetaMathQA dataset based on the Llemma-7B foundation model, significantly improving mathematical problem-solving capabilities.
Large Language Model
Transformers

M
meta-math
3,192
16
Metamath 7B V1.0
MetaMath-Llemma-7B is a mathematical reasoning model fine-tuned on the MetaMathQA dataset, demonstrating excellent performance on GSM8K and MATH datasets.
Large Language Model
Transformers

M
meta-math
278
27
Wizardmath 7B V1.0
WizardMath is a large language model empowered by Reinforced Learning from Evolved Instruction Feedback (RLEIF), specifically designed to enhance mathematical reasoning capabilities.
Large Language Model
Transformers

W
WizardLMTeam
653
54
Featured Recommended AI Models